407 research outputs found
Lessons from Building Acoustic Models with a Million Hours of Speech
This is a report of our lessons learned building acoustic models from 1
Million hours of unlabeled speech, while labeled speech is restricted to 7,000
hours. We employ student/teacher training on unlabeled data, helping scale out
target generation in comparison to confidence model based methods, which
require a decoder and a confidence model. To optimize storage and to
parallelize target generation, we store high valued logits from the teacher
model. Introducing the notion of scheduled learning, we interleave learning on
unlabeled and labeled data. To scale distributed training across a large number
of GPUs, we use BMUF with 64 GPUs, while performing sequence training only on
labeled data with gradient threshold compression SGD using 16 GPUs. Our
experiments show that extremely large amounts of data are indeed useful; with
little hyper-parameter tuning, we obtain relative WER improvements in the 10 to
20% range, with higher gains in noisier conditions.Comment: "Copyright 2019 IEEE. Personal use of this material is permitted.
Permission from IEEE must be obtained for all other uses, in any current or
future media, including reprinting/republishing this material for advertising
or promotional purposes, creating new collective works, for resale or
redistribution to servers or lists, or reuse of any copyrighted component of
this work in other works.
Study of gas density profile about a plastic test piece under vacuum
Nevada Shocker is a 540KV, 7O, 50ns, pulsed power device based on Marx Bank and Blumlein technologies. When this machine fires, the energy is funneled to a plastic insulator supported by two circular electrodes. The purpose is to study the surface breakdown properties of the plastic. It is hypothesized that a thin layer of gas exists between hard vacuum and the solid. The goal is to determine change in the gas profile and suggest how it may contribute to surface flashover; Stimulated desorption techniques such as photon stimulated desorption and thermal stimulated desorption are employed in characterizing the gas profile. The plastic under test is Rexolite. A vacuum system with Rexolite plastic is brought down to pressure levels of 10-8 to 10-9 Torr to study the presence of the thin layer of select gas molecules. A Nd: YAG pulsed laser is used to stimulate the surface of the plastic under vacuum
Garbage Collection for General Graphs
Garbage collection is moving from being a utility to a requirement of every modern programming language. With multi-core and distributed systems, most programs written recently are heavily multi-threaded and distributed. Distributed and multi-threaded programs are called concurrent programs. Manual memory management is cumbersome and difficult in concurrent programs. Concurrent programming is characterized by multiple independent processes/threads, communication between processes/threads, and uncertainty in the order of concurrent operations. The uncertainty in the order of operations makes manual memory management of concurrent programs difficult. A popular alternative to garbage collection in concurrent programs is to use smart pointers. Smart pointers can collect all garbage only if developer identifies cycles being created in the reference graph. Smart pointer usage does not guarantee protection from memory leaks unless cycle can be detected as process/thread create them. General garbage collectors, on the other hand, can avoid memory leaks, dangling pointers, and double deletion problems in any programming environment without help from the programmer. Concurrent programming is used in shared memory and distributed memory systems. State of the art shared memory systems use a single concurrent garbage collector thread that processes the reference graph. Distributed memory systems have very few complete garbage collection algorithms and those that exist use global barriers, are centralized and do not scale well. This thesis focuses on designing garbage collection algorithms for shared memory and distributed memory systems that satisfy the following properties: concurrent, parallel, scalable, localized (decentralized), low pause time, high promptness, no global synchronization, safe, complete, and operates in linear time
Implementing wireless local area network (WLAN) in United Technology (M) Sdn Bhd
Wireless becomes part and parcel in today's world. The widespread strategic reliance on
networking among competitive businesses and the meteoric growth of the internet and
online services are strong testimonies to the benefits of shared data and resources. With
wireless LAN, users can access shared information without looking for place to plug in
and network managers can set up of augment networks without installing or moving
wires. Wireless LANs offers productivity, service, and convenience and cost advantages
over traditional wired networks.
Thus, this piece of industrial training report is summarizing a complete scenario in setting
up a WLAN in United Technology (M) Sdn Bhd under a team of 6 personnel, whereby 1
am one of them. (Author's abstract
Unmasking Falsehoods in Reviews: An Exploration of NLP Techniques
In the contemporary digital landscape, online reviews have become an
indispensable tool for promoting products and services across various
businesses. Marketers, advertisers, and online businesses have found incentives
to create deceptive positive reviews for their products and negative reviews
for their competitors' offerings. As a result, the writing of deceptive reviews
has become an unavoidable practice for businesses seeking to promote themselves
or undermine their rivals. Detecting such deceptive reviews has become an
intense and ongoing area of research. This research paper proposes a machine
learning model to identify deceptive reviews, with a particular focus on
restaurants. This study delves into the performance of numerous experiments
conducted on a dataset of restaurant reviews known as the Deceptive Opinion
Spam Corpus. To accomplish this, an n-gram model and max features are developed
to effectively identify deceptive content, particularly focusing on fake
reviews. A benchmark study is undertaken to explore the performance of two
different feature extraction techniques, which are then coupled with five
distinct machine learning classification algorithms. The experimental results
reveal that the passive aggressive classifier stands out among the various
algorithms, showcasing the highest accuracy not only in text classification but
also in identifying fake reviews. Moreover, the research delves into data
augmentation and implements various deep learning techniques to further enhance
the process of detecting deceptive reviews. The findings shed light on the
efficacy of the proposed machine learning approach and offer valuable insights
into dealing with deceptive reviews in the realm of online businesses.Comment: 6 pages, 3 figure
Transfer Learning via Contextual Invariants for One-to-Many Cross-Domain Recommendation
The rapid proliferation of new users and items on the social web has
aggravated the gray-sheep user/long-tail item challenge in recommender systems.
Historically, cross-domain co-clustering methods have successfully leveraged
shared users and items across dense and sparse domains to improve inference
quality. However, they rely on shared rating data and cannot scale to multiple
sparse target domains (i.e., the one-to-many transfer setting). This, combined
with the increasing adoption of neural recommender architectures, motivates us
to develop scalable neural layer-transfer approaches for cross-domain learning.
Our key intuition is to guide neural collaborative filtering with
domain-invariant components shared across the dense and sparse domains,
improving the user and item representations learned in the sparse domains. We
leverage contextual invariances across domains to develop these shared modules,
and demonstrate that with user-item interaction context, we can learn-to-learn
informative representation spaces even with sparse interaction data. We show
the effectiveness and scalability of our approach on two public datasets and a
massive transaction dataset from Visa, a global payments technology company
(19% Item Recall, 3x faster vs. training separate models for each domain). Our
approach is applicable to both implicit and explicit feedback settings.Comment: SIGIR 202
- …